973 research outputs found
Weighted entropy and optimal portfolios for risk-averse Kelly investments
Following a series of works on capital growth investment, we analyse
log-optimal portfolios where the return evaluation includes `weights' of
different outcomes. The results are twofold: (A) under certain conditions, the
logarithmic growth rate leads to a supermartingale, and (B) the optimal
(martingale) investment strategy is a proportional betting. We focus on
properties of the optimal portfolios and discuss a number of simple examples
extending the well-known Kelly betting scheme.
An important restriction is that the investment does not exceed the current
capital value and allows the trader to cover the worst possible losses.
The paper deals with a class of discrete-time models. A continuous-time
extension is a topic of an ongoing study
A variational principle for cyclic polygons with prescribed edge lengths
We provide a new proof of the elementary geometric theorem on the existence
and uniqueness of cyclic polygons with prescribed side lengths. The proof is
based on a variational principle involving the central angles of the polygon as
variables. The uniqueness follows from the concavity of the target function.
The existence proof relies on a fundamental inequality of information theory.
We also provide proofs for the corresponding theorems of spherical and
hyperbolic geometry (and, as a byproduct, in spacetime). The spherical
theorem is reduced to the euclidean one. The proof of the hyperbolic theorem
treats three cases separately: Only the case of polygons inscribed in compact
circles can be reduced to the euclidean theorem. For the other two cases,
polygons inscribed in horocycles and hypercycles, we provide separate
arguments. The hypercycle case also proves the theorem for "cyclic" polygons in
spacetime.Comment: 18 pages, 6 figures. v2: typos corrected, final versio
Maximum Entropy Linear Manifold for Learning Discriminative Low-dimensional Representation
Representation learning is currently a very hot topic in modern machine
learning, mostly due to the great success of the deep learning methods. In
particular low-dimensional representation which discriminates classes can not
only enhance the classification procedure, but also make it faster, while
contrary to the high-dimensional embeddings can be efficiently used for visual
based exploratory data analysis.
In this paper we propose Maximum Entropy Linear Manifold (MELM), a
multidimensional generalization of Multithreshold Entropy Linear Classifier
model which is able to find a low-dimensional linear data projection maximizing
discriminativeness of projected classes. As a result we obtain a linear
embedding which can be used for classification, class aware dimensionality
reduction and data visualization. MELM provides highly discriminative 2D
projections of the data which can be used as a method for constructing robust
classifiers.
We provide both empirical evaluation as well as some interesting theoretical
properties of our objective function such us scale and affine transformation
invariance, connections with PCA and bounding of the expected balanced accuracy
error.Comment: submitted to ECMLPKDD 201
Relay Backpropagation for Effective Learning of Deep Convolutional Neural Networks
Learning deeper convolutional neural networks becomes a tendency in recent
years. However, many empirical evidences suggest that performance improvement
cannot be gained by simply stacking more layers. In this paper, we consider the
issue from an information theoretical perspective, and propose a novel method
Relay Backpropagation, that encourages the propagation of effective information
through the network in training stage. By virtue of the method, we achieved the
first place in ILSVRC 2015 Scene Classification Challenge. Extensive
experiments on two challenging large scale datasets demonstrate the
effectiveness of our method is not restricted to a specific dataset or network
architecture. Our models will be available to the research community later.Comment: Technical report for our submissions to the ILSVRC 2015 Scene
Classification Challenge, where we won the first plac
Information complexity of the AND function in the two-Party, and multiparty settings
In a recent breakthrough paper [M. Braverman, A. Garg, D. Pankratov, and O.
Weinstein, From information to exact communication, STOC'13] Braverman et al.
developed a local characterization for the zero-error information complexity in
the two party model, and used it to compute the exact internal and external
information complexity of the 2-bit AND function, which was then applied to
determine the exact asymptotic of randomized communication complexity of the
set disjointness problem.
In this article, we extend their results on AND function to the multi-party
number-in-hand model by proving that the generalization of their protocol has
optimal internal and external information cost for certain distributions. Our
proof has new components, and in particular it fixes some minor gaps in the
proof of Braverman et al
Competitive portfolio selection using stochastic predictions
We study a portfolio selection problem where a player attempts to maximise a utility function that represents the growth rate of wealth. We show that, given some stochastic predictions of the asset prices in the next time step, a sublinear expected regret is attainable against an optimal greedy algorithm, subject to tradeoff against the \accuracy" of such predictions that learn (or improve) over time. We also study the effects of introducing transaction costs into the model
Scanner Invariant Representations for Diffusion MRI Harmonization
Purpose: In the present work we describe the correction of diffusion-weighted
MRI for site and scanner biases using a novel method based on invariant
representation.
Theory and Methods: Pooled imaging data from multiple sources are subject to
variation between the sources. Correcting for these biases has become very
important as imaging studies increase in size and multi-site cases become more
common. We propose learning an intermediate representation invariant to
site/protocol variables, a technique adapted from information theory-based
algorithmic fairness; by leveraging the data processing inequality, such a
representation can then be used to create an image reconstruction that is
uninformative of its original source, yet still faithful to underlying
structures. To implement this, we use a deep learning method based on
variational auto-encoders (VAE) to construct scanner invariant encodings of the
imaging data.
Results: To evaluate our method, we use training data from the 2018 MICCAI
Computational Diffusion MRI (CDMRI) Challenge Harmonization dataset. Our
proposed method shows improvements on independent test data relative to a
recently published baseline method on each subtask, mapping data from three
different scanning contexts to and from one separate target scanning context.
Conclusion: As imaging studies continue to grow, the use of pooled multi-site
imaging will similarly increase. Invariant representation presents a strong
candidate for the harmonization of these data
Physics‐constrained non‐Gaussian probabilistic learning on manifolds
International audienceAn extension of the probabilistic learning on manifolds (PLoM), recently introduced by the authors, has been presented: In addition to the initial data set given for performing the probabilistic learning, constraints are given, which correspond to statistics of experiments or of physical models. We consider a non-Gaussian random vector whose unknown probability distribution has to satisfy constraints. The method consists in constructing a generator using the PLoM and the classical Kullback-Leibler minimum cross-entropy principle. The resulting optimization problem is reformulated using Lagrange multipliers associated with the constraints. The optimal solution of the Lagrange multipliers is computed using an efficient iterative algorithm. At each iteration, the Markov chainMonte Carlo algorithm developed for the PLoM is used, consisting in solving an Itô stochastic differential equation that is projected on a diffusion-maps basis. The method and the algorithm are efficient and allow the construction ofprobabilistic models for high-dimensional problems from small initial data sets and for which an arbitrary number of constraints are specified. The first application is sufficiently simple in order to be easily reproduced. The second one is relative to a stochastic elliptic boundary value problem in high dimension
End-to-End Joint Antenna Selection Strategy and Distributed Compress and Forward Strategy for Relay Channels
Multi-hop relay channels use multiple relay stages, each with multiple relay
nodes, to facilitate communication between a source and destination.
Previously, distributed space-time codes were proposed to maximize the
achievable diversity-multiplexing tradeoff, however, they fail to achieve all
the points of the optimal diversity-multiplexing tradeoff. In the presence of a
low-rate feedback link from the destination to each relay stage and the source,
this paper proposes an end-to-end antenna selection (EEAS) strategy as an
alternative to distributed space-time codes. The EEAS strategy uses a subset of
antennas of each relay stage for transmission of the source signal to the
destination with amplify and forwarding at each relay stage. The subsets are
chosen such that they maximize the end-to-end mutual information at the
destination. The EEAS strategy achieves the corner points of the optimal
diversity-multiplexing tradeoff (corresponding to maximum diversity gain and
maximum multiplexing gain) and achieves better diversity gain at intermediate
values of multiplexing gain, versus the best known distributed space-time
coding strategies. A distributed compress and forward (CF) strategy is also
proposed to achieve all points of the optimal diversity-multiplexing tradeoff
for a two-hop relay channel with multiple relay nodes.Comment: Accepted for publication in the special issue on cooperative
communication in the Eurasip Journal on Wireless Communication and Networkin
- …